LogoNotes by Cole Gawin
v0

Goals of LING 385

  • learn the ideas behind the computations that have allowed modern language technologies to become important parts of society
    • and the practical computations that implement the ideas
  • learn how scientific ideas from different centuries are creatively put together to make new seemingly revolutionary ideas
  • learn about how commonly used tools actually work

Language, Technology, and Society

  • great deal of what we hear about language technologies is the harm they can cause
    • very valid concerns
  • however, technologies can also be used for good
  • understanding how these systems work and acting on this understanding is the best way to make sure they are used for good causes

Language Technology: how much do we know?

  • most current systems are open-source
    • we know every mathematical idea and algorithm powering them
  • despite this, we don't really know how it all comes together on the lowest level

Neural Networks: an Interdisciplinary Discipline

  • interest in NNs waned till the physicists got interested in late 70’s
  • great deal of physics is about figuring out the macroscopic properties of systems (e.g., gases, polymers, magnets) from the microscopic laws governing atoms and molecules
  • Hopfield and Smolensky saw the analogy of molecules to neurons and macroscopic material properties to cognitive processes
  • in 1986, when Rumelhart and McClelland published Parallel Distributed Processing
    • brought together work by psychologists, computer scientists, and physicists
  • after that, NN research waned again
  • in 2006, deep learning (using neural networks in industry) was born
  • now, people have started trying to understand how these systems work by going back to Hopfield, Rumelhart and McClelland

Associative Memory

  • based on associations of two things (or between part or whole)
  • John Hopfield, in 1982, showed that some very simple calculations, analogous to ones in physics, can model associative memory
    • 40 years later, his theory was re-used, 2021–current, to try to figure out how systems like ChatGPT actually work
  • heteroassociative memory = remembering some piece of information from another
    • DALL-E is heteroassociative (text → image)
  • auto-associative memory = remembering a whole from a part, or a noisy version of the whole
    • ChatGPT is auto-associative (text → text)